Summrize Logo

Best Books About Artificial Intelligence

Best Books Lists

The authors present a framework for understanding the role of rules and standard operating procedures (SOPs) in organizations:

  1. SOPs reduce individual cognitive load by providing pre-specified actions
  2. SOPs enable reliability and predictability across an organization by ensuring different people/groups take consistent, coordinated actions without extensive communication
  3. SOPs "glue" different parts of an organization together in an interdependent system resistant to change
  4. Replacing SOPs with AI-enhanced decisions reduces reliability and predictability, unsticking the organizational glue
  5. Transforming rules into AI-enhanced decisions often requires systemic change to re-establish coordination in new ways

Section: 2, Chapter: 6

The authors argue that AI presents an "innovator's dilemma" that makes it difficult for incumbent firms to adopt the technology in a timely and effective manner. The key challenges:

  • Incumbent firms are optimized for their existing systems and metrics, while AI often requires new systems and metrics to deliver value
  • Adopting AI may cannibalize incumbents' existing profitable businesses, while startups have no such conflicts
  • Incumbents' organizational structures and processes are "glued" together in ways that resist the changes required to fully exploit AI

AI systems often decouple prediction from judgment in ways that disrupt incumbents' existing decision-making structures and power dynamics As a result, the authors expect AI to be more rapidly adopted by startups than incumbents, leading to significant disruption as new entrants scale up AI-powered systems that incumbents struggle to match.

Section: 4, Chapter: 10

The authors argue that AI bias is often a reflection of flaws in the underlying system rather than a problem with the AI algorithm itself. Key points:

  • AI predictions are only as good as the data used to train the algorithm; if that data reflects historical biases or discrimination, the predictions will too
  • Fixing AI bias requires not just tweaking the algorithm but also addressing the root causes of bias in the data-generating process
  • In many cases, AI can be a powerful tool for detecting and mitigating bias, by making it visible and quantifiable in a way that human decision-making does not
  • The key is to have a clear definition of fairness and a systematic process for auditing AI systems against that definition

While imperfect, the authors argue AI is often less biased than human decision-makers and offers greater potential for bias detection and mitigation. The challenge is to design systems that realize that potential.

Section: 6, Chapter: 18

When deploying AI systems, it's critical to ensure the predictions are valid for the decisions you want to make. Some key considerations:

  • Distinguish between correlation and causation. AI predictions based on historical data may identify correlations that don't reflect causal relationships.
  • Collect data that covers the full range of relevant situations. Predicting outside the "support of your data" is risky.
  • Where possible, use randomized experiments to collect data that reliably measures causal impact. Leading tech companies now employ many economists and statisticians focused on causal inference.

Section: 1, Chapter: 3

The authors argue that the predominant "task-level" paradigm for thinking about AI adoption and impact is misguided and limiting. The key points:

  • Most leaders and experts focus on identifying specific tasks that AI could perform better than humans and calculating the labor substitution effects
  • However, the greatest value from AI comes not from piecemeal task substitution but from reimagining entire systems and processes around AI capabilities
  • Focusing on tasks leads to small-scale point solutions, while system-level thinking enables transformative new structures and strategies
  • The biggest AI successes to date, like Netflix, Amazon, and YouTube, have come from system-level innovation rather than task substitution

Leaders should adopt a "system mindset" and focus on how AI predictions could enable fundamentally new approaches to delivering value, even if those approaches are inferior on traditional metrics.

Section: 3, Chapter: 8

When deploying AI systems, it's critical to remember that machines don't actually make decisions, humans do.

  • AI systems automate the execution of decisions, but the judgment embedded in those systems always comes from humans
  • The humans who specify the objectives, metrics, and reward functions for an AI system are the real decision-makers, not the system itself
  • Responsible AI deployment requires carefully specifying the judgment an AI system should execute, anticipating edge cases, and monitoring for unexpected behaviors

Leaders should ensure clear human accountability for the decisions executed by AI systems, rather than allowing those systems to be treated as autonomous agents.

Section: 4, Chapter: 11

The authors illustrate AI's impact on the allocation of power with the case of the Flint, Michigan water crisis.

  • City officials initially ignored data showing high lead levels in the water supply, relying on flawed testing and flawed judgment
  • Outside researchers used AI to predict which homes were likely to have lead pipes and successfully pressured officials to target remediation efforts based on their predictions
  • The researchers' models were technically superior to the city's testing methods, but the key factor was that the researchers had better judgment about how to act on the predictions
  • Ultimately, power over the response shifted from city officials to a court-appointed monitor, who had the authority to override officials' flawed judgment and act on the researchers' predictions The case illustrates how AI can shift decision-making power to those with superior judgment, even if they don't have the best predictions or the formal authority.

Section: 5, Chapter: 15

Adopting AI requires a shift from deterministic to probabilistic thinking in decision-making. Some key mindset changes:

  • Embrace uncertainty and accept that all predictions are probabilistic, rather than expecting definitive answers
  • Think in terms of expected value, weighing the probability of different outcomes rather than trying to eliminate all risk
  • Be transparent about the confidence level of predictions and the potential for error, rather than presenting predictions as certain
  • Build processes to periodically retrain models and update predictions as new data becomes available
  • Develop ethical frameworks and oversight mechanisms to ensure predictions are applied with appropriate human judgment

Section: 5, Chapter: 14

"Like SOPs, checklists are the manifestation of rules and the need to follow them. They are there to ensure reliability and reduce error. The alternative is that people make decisions based on their own observations. While switching from a rule to a decision may improve the quality of that particular action, it may also create problems and uncertainty for other people."

Section: 2, Chapter: 6

The authors argue that AI's most profound impact may be on the process of innovation and invention itself. Key points:

  • AI enables faster and cheaper hypothesis generation and testing, accelerating the innovation cycle
  • AI-powered tools like AlphaFold are dramatically reducing the time and cost of key innovation steps like protein structure prediction
  • Just as previous research tools like microscopes enabled the discovery of the germ theory of disease, AI is a "new method of invention" that will have cascading effects on multiple domains

Section: 3, Chapter: 9

The authors explain the difference between decisions and rules:

  • Decisions allow you to take into account situational information but incur cognitive costs to gather information and deliberate
  • Rules avoid cognitive costs but result in the same action regardless of the situation The key factors that determine if the cognitive cost of a decision is worthwhile are:
  1. The consequences of the decision - more consequential decisions are worth deciding
  2. The cost of information - if information is cheap, decisions become more attractive

Section: 2, Chapter: 4

The authors argue that the essence of recent advances in AI is that they represent a dramatic improvement in prediction - the ability to take information you have and generate information you don't have. Prediction is a key input into decision making. As prediction becomes cheaper, we will use more of it and the value of other inputs to decision making like human prediction will fall while the value of complements like data and judgment will rise. Judgment is determining the relative payoff or reward to different actions - it is a statement of what we want, while prediction tells us the likelihood of different outcomes.

Section: 1, Chapter: 3

"Predicting on the support of your data is not as simple as collecting data from a wider variety of settings to ensure you aren't extrapolating too much or avoiding predicting too far into the future. Sometimes the data you need doesn't exist. This underlies the refrain repeated in every statistics course worldwide: correlation is not necessarily causation."

Section: 1, Chapter: 3

As a leader, you should examine the existing rules and standard operating procedures in your organization and evaluate if they can be turned into decisions enhanced by AI predictions. Look for situations where:

  • The rule leads to costly errors in some situations that could be mitigated by deciding differently
  • AI can provide cheap, high-quality predictions to enable better situational decisions
  • The cost savings or performance gains from better decisions justify the cognitive cost and reduced reliability of turning the rule into an AI-informed decision

Section: 2, Chapter: 4

AI is increasing the returns to scale and scope in many industries, creating opportunities for companies to build significant competitive advantages. Some key strategies:

  • Invest in proprietary data assets that enable your AI systems to make better predictions than rivals
  • Build feedback loops that allow your AI systems to learn and improve over time, increasing their predictive edge
  • Look for opportunities to apply your AI-powered predictions in adjacent markets and product categories
  • Exploit the "flywheel effect" by using AI to attract more users/customers, generating more data, leading to better predictions, in a virtuous cycle

However, be aware that these same dynamics can help rivals build insurmountable leads if they get ahead in the AI race. In industries with strong AI feedback loops, being a fast follower may not be a viable strategy.

Section: 4, Chapter: 12

The authors use the example of modern airport design to illustrate the concept of "hidden uncertainty." Frequent air travelers arrive at the airport much earlier than their flights to accommodate the uncertainty around traffic, parking, security lines, flight delays, etc. Airports like Incheon Airport in South Korea now provide extensive amenities like spas, museums, gardens, and ice skating to make the inevitably long wait times more palatable.

However, this expensive infrastructure accommodates the hidden uncertainty rather than resolving it. The authors suggest that AI prediction could reduce the uncertainty and enable a new, more efficient equilibrium.

Section: 2, Chapter: 5

When adopting AI to optimize different parts of an interdependent system, beware of the "AI Bullwhip effect" when the output of one AI-optimized component becomes the input to another in an unanticipated way. For example:

  • A restaurant adopts an AI system to predict food orders and optimize its inventory, reducing waste
  • This makes the restaurant's orders to suppliers less predictable, forcing suppliers to carry more inventory
  • The added volatility ripples through the supply chain, forcing each tier to adopt its own

AI optimization Leaders should carefully map out interdependencies in their systems and plan for the impact of AI optimization on upstream and downstream components.

Section: 3, Chapter: 8

The authors discuss how AI could enable a transformation of the education system from the current "factory model" where students progress based on age to a personalized model where each student receives customized instruction based on their individual learning needs and pace. Key points:

  • In the factory model, the curriculum is tied to the student's age and grade rather than their individual progress, and teachers deliver one-size-fits-all instruction
  • AI-powered adaptive learning systems can predict the optimal next lesson for each student based on their performance, enabling them to progress at their own pace
  • Realizing this vision requires not just better AI but a redesign of the education system, including changes to student grouping, pedagogy, teacher training, facilities, etc.

Section: 2, Chapter: 6

The authors argue that one of AI's most profound effects is to decouple prediction, which is increasingly done by machines, from judgment, which remains the province of humans. Key implications:

  • Prediction and judgment are the two key ingredients of decision-making; before AI, they were tightly coupled in the minds of human decision-makers
  • AI allows prediction to be centralized and automated while judgment remains decentralized and human-driven
  • This decoupling creates opportunities to reimagine systems and processes, but also creates challenges around aligning predictions with appropriate judgment
  • As AI takes over more prediction, the key differentiator for human decision-makers will be their judgment, i.e. their ability to specify what objectives should be maximized

Section: 5, Chapter: 13

"Once the AI provides the prediction, then the people with the best judgment can shine... Once the AI provides the prediction, new systems can arise to take advantage of better, faster, and cheaper predictions and more appropriate judgment. In Prediction Machines, we highlighted an opportunity for Amazon to change its business model so that it ships items to your door before you even order. That business model now exists. Stitch Fix does it for clothes."

Section: 5, Chapter: 13

Airport operators should be wary of the disruptive potential of AI-powered navigation apps like Waze and Google Maps. Key considerations:

  • These apps can provide increasingly accurate, personalized predictions of travel time to the airport, reducing the need for passengers to budget large uncertainty buffers
  • As passengers become more confident in "just in time" airport arrival, demand for in-terminal retail and dining may fall significantly
  • Airport operators should explore ways to actively partner with navigation apps to shape behavior and preserve retail revenues, rather than being passive victims of disruption

Section: 2, Chapter: 5

The authors describe how AI is enabling innovators to test ideas via simulation rather than costly real-world experiments.

  • Pharmaceutical companies can use AI to predict the outcomes of clinical trials, enabling them to prioritize the most promising drug candidates
  • Aerospace companies can use AI-powered simulations to test new aircraft designs without building physical prototypes
  • E-commerce companies can use AI to simulate the impact of website changes on customer behavior before deploying them live

By making experimentation faster and cheaper, AI simulations accelerate innovation and reduce risk. However, moving too quickly from simulation to the real world can be dangerous, as the fatal accidents involving Boeing's 737 Max and Tesla's self-driving systems illustrate.

Section: 3, Chapter: 9

The authors present an "AI Canvas" framework for mapping how AI predictions could enable businesses and organizations to redesign their decision-making systems. The key steps:

  1. Define the core objective or "north star" of your organization
  2. Identify the key decisions required to achieve that objective, assuming the availability of perfect predictions
  3. For each decision, specify the prediction required as an input and the judgment required to act on that prediction
  4. Analyze how the decision-making roles and processes would need to change to incorporate AI predictions
  5. Redesign the overall system to maximize the value of AI predictions while preserving necessary human judgment

The authors apply the framework to an extended case study of the insurance industry, showing how AI could transform insurers' decision-making from underwriting and claims processing to customer acquisition and retention.

Section: 6, Chapter: 17

In 2016, a ProPublica investigation into the COMPAS criminal risk assessment tool concluded the tool was biased against Black defendants. Their analysis found that Black defendants who did not reoffend were 2x more likely to be classified as high-risk compared to White defendants.

The makers of COMPAS, Northpointe, countered that the model was equally accurate for White and Black defendants and had the same false positive rates for each risk score level, so could not be biased.

This sparked a heated debate in the algorithmic fairness community. A series of academic papers showed that the two notions of fairness - equal false positive rates and equal accuracy across groups - are mathematically incompatible if the base rates of the predicted variable differ across groups.

The COMPAS debate crystallized the realization that there are multiple conceptions of algorithmic fairness that often cannot be simultaneously satisfied. It brought the issue into the public eye and kickstarted the field of fairness in machine learning.

Section: 1, Chapter: 2

A striking example of the power of "intrinsic motivation" systems in AI is the case of Montezuma's Revenge, an Atari game that proved frustratingly difficult for standard reinforcement learning agents to solve.

The game requires extensive exploration to find sparse rewards, which is infeasible for agents only motivated by the explicit game score. By contrast, agents imbued with "artificial curiosity" - receiving intrinsic reward for discovering novel states or situations that surprise their worldview - are able to systematically explore the game world and uncover success.

Other examples:

  • The "NoveltyNet" agent developed by Bellemare and colleagues at DeepMind generated an intrinsic reward proportional to how unfamiliar a new game state was based on its experience. Seeking out these novel states allowed it to discover 15 of the 24 rooms in Montezuma's Revenge without relying on the game score.
  • Pathak and colleagues at Berkeley trained agents with an "Intrinsic Curiosity Module" that was rewarded for discovering states that surprised a neural network tasked with predicting the consequence of actions. This surprise-seeking agent achieved superhuman performance on many games.

So formulating a drive to discover novelty and resolve uncertainty proved to be a powerful substitute for extrinsic rewards in motivating learning and exploration. This echoes the curiosity-driven learning of infants and illustrates a key alternative mechanism to "classical" external reinforcement.

Section: 2, Chapter: 6

Chapter 7 explores how imitation learning - having machines learn by observing and mimicking human behavior - is both a distinctively human capability and a promising approach to building flexible AI systems.

  • Humans are unique in our ability and proclivity to imitate, which is a foundation of our intelligence. Even infants just a few days old can mimic facial expressions.
  • Imitation is powerful because it allows learning from a small number of expert demonstrations rather than extensive trial-and-error. It also enables learning unspoken goals and intangible skills.
  • Techniques like inverse reinforcement learning infer reward functions from examples of expert behavior, enabling machines to adopt the goals and values implicit in the demonstrated actions.
  • Imperfect imitation that captures the demonstrator's underlying intent can actually produce behavior that surpasses that of the teacher. This "value alignment" may be essential for building beneficial AI systems.
  • But imitation also has pitfalls - it tends to break down when the imitator has different capabilities than the demonstrator, or encounters novel situations. So imitation is powerful, but no panacea.

The big picture is that imitation learning is a distinctively human form of intelligence that is also a promising path to more human-compatible AI systems. But it must be thoughtfully combined with other forms of learning and adaptation to achieve robust real-world performance.

Section: 3, Chapter: 7

Chapter 3 makes the provocative case that often the most accurate models are the simplest ones, not complex neural networks, if the input features are wisely chosen.

Psychologist Paul Meehl showed in the 1950s that very simple statistical models consistently matched or beat expert human judgment at predicting things like academic performance or recidivism risk. Later work by Robyn Dawes in the 1970s demonstrated that even models with random feature weights (as long as they are positive) are highly competitive with human experts.

The key insight is that the predictive power comes from astute selection of the input features, not complex combinations of them. The experts' true skill is "knowing what to look for," then simple addition of those features does the rest.

This has major implications for model transparency. Wherever possible, simple, inspectable models should be preferred. And we should be extremely thoughtful about what features we choose to include since they, more than anything, drive the model's behavior.

Section: 1, Chapter: 3

Actionable insights for AI developers:

  • Make AI systems' confidence scores actually reflect statistical uncertainty, not just relative ranking
  • Build pipelines for "uncertainty handoff" to human oversight in high-stakes applications
  • Extensively test AI systems on out-of-distribution and adversarial inputs to probe overconfidence
  • Favor objectives and learning procedures that are robust to uncertainty over brittle "point estimates"

The upshot is that well-calibrated uncertainty is a feature, not a bug, for AI systems operating in the open world. We should invest heavily in uncertainty estimation techniques and make them a core component of AI system design.

Section: 3, Chapter: 9

“As we’re on the cusp of using machine learning for rendering basically all kinds of consequential decisions about human beings in domains such as education, employment, advertising, health care and policing, it is important to understand why machine learning is not, by default, fair or just in any meaningful way.”

Section: 1, Chapter: 2

A cautionary tale about the limits of imitation learning comes from the experience of UC Berkeley researchers in using human gameplay data to train AI agents to play the game Montezuma's Revenge.

The game is notoriously difficult for standard reinforcement learning agents due to sparse and deceptive rewards. So the researchers tried "jumpstarting" the agent's learning by pre-training it to mimic human players based on YouTube videos of successful playthroughs.

This worked to an extent - the imitation-bootstrapped agent made more progress than any previous learning agent. But it also ran into problems:

  • The human videos showed successful runs, not the many failed attempts. So the agent never saw recoveries from mistakes and couldn't replicate them.
  • The agent lacked the humans' general world knowledge, so it interpreted their actions overly literally. E.g. it learned to mimic a player's aimless "victory dance" after completing a level instead of moving on.
  • Mimicry couldn't account for differences in reaction speed and control precision between humans and the AI. The agent needed to develop its own robust behaviors.

Eventually, DeepMind researchers found that "intrinsic motivation" approaches were more successful on Montezuma's Revenge than imitation learning. The game illustrates how one-shot mimicry of experts is no substitute for flexible trial-and-error learning and adaptation. Imitation is most powerful when combined with other learning mechanisms to overcome its blind spots.

Section: 3, Chapter: 7

Some key aspects of practical IRL frameworks:

  • Accounting for expert suboptimality/imperfection
  • Allowing for reward functions more complex than linear combinations of pre-defined features
  • Admitting reward ambiguity (many reward functions can explain a given policy)
  • Leveraging interactivity and active learning to efficiently narrow down reward functions

IRL is not a complete solution to AI value alignment, but a powerful conceptual and algorithmic tool. It provides a principled way to specify objectives for AI systems by demonstration and example. And it forces us to grapple with the difficulty of distilling clear "reward functions" from human behavior.

Section: 3, Chapter: 8

Chapter 1 explores how bias and unfairness in machine learning models frequently stems from the data used to train them not being representative of the real world. Some key examples:

  • Face recognition systems performing poorly on Black faces because their training data contained mostly White faces
  • Word embedding models picking up on gender stereotypes because those associations were present in the large corpora of human-generated text used to train them
  • Amazon's resume screening tool downranking women because it was trained on past resumes, which skewed male

The overarching lesson is that a model is only as unbiased as the data it learns from. Careful attention needs to be paid to the composition of training datasets to ensure they are adequately representative of the real-world populations the models will be applied to. There are also techniques to try to debias models, like identifying and removing stereotyped associations, but starting with representative data is the first line of defense against bias.

Section: 1, Chapter: 1

    Summrize Footer

    Subscribe to Our Newsletter

    Become smarter every day with key takeaways delivered straight to your inbox. Perfect for busy people who want to learn from the smartest minds in just minutes!